high-performance computing
Powering HPC with next-generation CPUs
From genomics to AI, CPUs continue to lead high-performance computing by offering unmatched flexibility, reliability, and scale, says Microsoft Azure's Evan Burness and Intersect360 Research's Addison Snell. For all the excitement around GPUs--the workhorses of today's AI revolution--the central processing unit (CPU) remains the backbone of high-performance computing (HPC). CPUs still handle 80% to 90% of HPC workloads globally, powering everything from climate modeling to semiconductor design. Far from being eclipsed, they're evolving in ways that make them more competitive, flexible, and indispensable than ever. The competitive landscape around CPUs has intensified. Once dominated almost exclusively by Intel's x86 chips, the market now includes powerful alternatives based on ARM and even emerging architectures like RISC-V.
- North America > United States > Massachusetts (0.05)
- Asia > Japan (0.05)
The supercomputer set to supercharge America's AI future
A growing number of fire departments across the country are turning to artificial intelligence to help detect and respond to wildfires more quickly. A major breakthrough in artificial intelligence and high-performance computing is on the way, and it's coming from Georgia Tech. Backed by a 20 million investment from the National Science Foundation (NSF), the university is building a supercomputer named Nexus. It's expected go online in spring 2026. Sign up for my FREE CyberGuy Report Get my best tech tips, urgent security alerts and exclusive deals delivered straight to your inbox.
- Media > News (0.32)
- Health & Medicine > Pharmaceuticals & Biotechnology (0.31)
MARCO: Multi-Agent Code Optimization with Real-Time Knowledge Integration for High-Performance Computing
Rahman, Asif, Cvetkovic, Veljko, Reece, Kathleen, Walters, Aidan, Hassan, Yasir, Tummeti, Aneesh, Torres, Bryan, Cooney, Denise, Ellis, Margaret, Nikolopoulos, Dimitrios S.
--Large language models (LLMs) have transformed software development through code generation capabilities, yet their effectiveness for high-performance computing (HPC) remains limited. HPC code requires specialized optimizations for parallelism, memory efficiency, and architecture-specific considerations that general-purpose LLMs often overlook. We present MARCO (Multi-Agent Reactive Code Optimizer), a novel framework that enhances LLM-generated code for HPC through a specialized multi-agent architecture. MARCO employs separate agents for code generation and performance evaluation, connected by a feedback loop that progressively refines optimizations. A key innovation is MARCO's web-search component that retrieves real-time optimization techniques from recent conference proceedings and research publications, bridging the knowledge gap in pre-trained LLMs. Our extensive evaluation on the LeetCode 75 problem set demonstrates that MARCO achieves a 14.6% average runtime reduction compared to Claude 3.5 Sonnet alone, while the integration of the web-search component yields a 30.9% performance improvement over the base MARCO system. These results highlight the potential of multi-agent systems to address the specialized requirements of high-performance code generation, offering a cost-effective alternative to domain-specific model fine-tuning. High-performance computing (HPC) represents the pinnacle of computational power, utilizing clusters of computing resources to overcome the limitations of individual machines. HPC's fundamental advantage lies in implementing parallel processing techniques that maximize processor cluster performance, enabling complex data processing and mathematical calculations that would otherwise be infeasible [33]. HPC has been instrumental in driving innovation across diverse domains including climate modeling, astrophysics simulations, pharmaceutical research, energy optimization, financial risk analysis, and training state-of-the-art machine learning models, particularly Large Language Models [7, 11, 15, 20, 30, 47].
- North America > United States > Virginia (0.04)
- Europe > Switzerland (0.04)
- Asia > Middle East > Jordan (0.04)
- Banking & Finance (0.48)
- Information Technology > Software (0.48)
Design optimization for high-performance computing using FPGA
Isik, Murat, Inadagbo, Kayode, Aktas, Hakan
Reconfigurable architectures like Field Programmable Gate Arrays (FPGAs) have been used for accelerating computations in several domains because of their unique combination of flexibility, performance, and power efficiency. However, FPGAs have not been widely used for high-performance computing, primarily because of their programming complexity and difficulties in optimizing performance. We optimize Tensil AI's open-source inference accelerator for maximum performance using ResNet20 trained on CIFAR in this paper in order to gain insight into the use of FPGAs for high-performance computing. In this paper, we show how improving hardware design, using Xilinx Ultra RAM, and using advanced compiler strategies can lead to improved inference performance. We also demonstrate that running the CIFAR test data set shows very little accuracy drop when rounding down from the original 32-bit floating point. The heterogeneous computing model in our platform allows us to achieve a frame rate of 293.58 frames per second (FPS) and a %90 accuracy on a ResNet20 trained using CIFAR. The experimental results show that the proposed accelerator achieves a throughput of 21.12 Giga-Operations Per Second (GOP/s) with a 5.21 W on-chip power consumption at 100 MHz. The comparison results with off-the-shelf devices and recent state-of-the-art implementations illustrate that the proposed accelerator has obvious advantages in terms of energy efficiency.
- North America > United States > New York (0.04)
- Asia > Middle East > Republic of Türkiye > Nigde Province > Nigde (0.04)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.04)
- Semiconductors & Electronics (0.70)
- Information Technology (0.46)
- Banking & Finance (0.46)
How Data Centers are enabling Artificial Intelligence (AI) - Dgtl Infra
The rapid growth of data generation fueled by artificial intelligence (AI) has transformed how data is stored, processed, managed, and transferred, while increasing the demand for computing power across cloud and edge data centers. To meet the demand generated by AI, data centers are evolving and adapting their design, power infrastructure, and cooling equipment in various unique ways. Data centers provide vast computing resources and storage, enabling artificial intelligence (AI) to process massive datasets for training and inference. By hosting specialized hardware such as GPUs and TPUs, data centers accelerate complex calculations, supporting AI applications and workloads. As Dgtl Infra delves deeper into the evolving relationship between artificial intelligence and data centers, we offer insights on power consumption, cooling requirements, and the pivotal role of data centers in supporting AI.
NVIDIA-Knowing the power of Artificial Intelligence
NVIDIA is a technology company that specializes in the design and manufacturing of graphics processing units (GPUs) and system on a chip (SoC) units for the gaming and professional markets. The company was founded in 1993 and is headquartered in Santa Clara, California. NVIDIA's GPUs are widely used in gaming PCs, workstations, and data centers for tasks such as machine learning, data analytics, and scientific simulations. The company also designs and manufactures other products such as Tegra mobile processors, Jetson embedded platform and CUDA parallel computing platform. The company's products are based on its proprietary CUDA architecture, which allows for parallel processing of large amounts of data.
Azure high-performance computing powers energy industry innovation
Azure high-performance computing provides a platform for energy industry innovation at scale. Global energy demand has rapidly increased over the last few years and looks set to continue accelerating at such a pace. With a booming middle class, economic growth, digitization, urbanization, and increased mobility of populations, energy suppliers are in a race to leverage the development of new technologies that can more optimally and sustainably generate, store, and transport energy to consumers. With the impact of climate change adding urgency to minimizing energy waste, in addition to optimizing power production leaders in the renewable energy as well as oil and gas industries are accelerating sector-wide innovation initiatives that can drive differentiated impact and outcomes at scale. As the population of developing countries continues to expand, the energy needs of billions of additional people in rural and especially urban areas will need to be catered to.
Remote Sensing
Remote sensing (RS) plays an important role gathering data in many critical domains (e.g., global climate change, risk assessment and vulnerability reduction of natural hazards, resilience of ecosystems, and urban planning). Retrieving, managing, and analyzing large amounts of RS imagery poses substantial challenges. Google Earth Engine (GEE) provides a scalable, cloud-based, geospatial retrieval and processing platform. GEE also provides access to the vast majority of freely available, public, multi-temporal RS data and offers free cloud-based computational power for geospatial data analysis. Artificial intelligence (AI) methods are a critical enabling technology to automating the interpretation of RS imagery, particularly on object-based domains, so the integration of AI methods into GEE represents a promising path towards operationalizing automated RS-based monitoring programs. In this article, we provide a systematic review of relevant literature to identify recent research that incorporates AI methods in GEE. We then discuss some of the major challenges of integrating GEE and AI and identify several priorities for future research. We developed an interactive web application designed to allow readers to intuitively and dynamically review the publications included in this literature review.
Altair Announces Digital Twin Solution
TROY, MI, Oct 4, 2022 – Altair, a global leader in computational science and artificial intelligence (AI), announced the launch of its broad digital twin solution that features the market's most connected, cross-functional capabilities that can be deployed through any and every stage of a product lifecycle. "Altair offers the market's premier digital twin solution that can transform the way people and organizations design, develop, implement, and improve products and processes," said Sam Mahalingam, chief technology officer, Altair. "Moving forward, we will continue establishing our digital twin leadership to provide further democratized, more accessible digital twin solutions." Combining Altair's leading simulation, high-performance computing (HPC), AI, data analytics, and Internet of Things (IoT) capabilities, companies can apply digital twin technology at any stage of the product lifecycle -- from concept through in-service -- as part of a cross-functional, enterprise-wide effort that advances collaboration and eliminates departmental silos. Additionally, Altair's open, vendor-agnostic digital twin solution is the premier offering that gives customers the flexibility to run Altair software anywhere – whether on-site, in the cloud, hybrid, or via plug-and-play appliances – and the freedom to choose from a comprehensive toolset through a cost-effective, units-based licensing model called Altair Units.
- Banking & Finance (0.33)
- Energy (0.32)
Graphcore IPUs adopted in Argonne National Lab's AI Testbed
Argonne National Laboratory, a multidisciplinary science and engineering research centre operated by the University of Chicago for the U.S. Department of Energy, has installed Graphcore's Intelligence Processing Units (IPUs) within its AI Testbed. The AI Testbed, operated by the Argonne Leadership Computing Facility (ALCF), enables researchers to explore next-generation machine learning applications and workloads to advance the use of AI for science. Technologies selected for the testbed, such as Graphcore's IPUs, complement the facility's current and next-generation supercomputers to provide a state-of-the-art environment that supports pioneering research at the intersection of AI, big data, and high-performance computing (HPC). The systems in the ALCF AI Testbed are purpose-built for machine learning and data-centric workloads, making them well suited to address challenges involving the increasingly large amounts of data produced by supercomputers, light sources, and particle accelerators among other powerful research tools. In addition, the testbed allows researchers to explore novel workflows that combine AI methods with simulation and experimental science to accelerate the pace of discovery.
- Energy (1.00)
- Government > Regional Government > North America Government > United States Government (0.79)